Goto

Collaborating Authors

 deep convolutional generative adversarial network


Statistical Physics of Deep Neural Networks: Generalization Capability, Beyond the Infinite Width, and Feature Learning

Ariosto, Sebastiano

arXiv.org Artificial Intelligence

Deep Neural Networks (DNNs) excel at many tasks, often rivaling or surpassing human performance. Yet their internal processes remain elusive, frequently described as "black boxes." While performance can be refined experimentally, achieving a fundamental grasp of their inner workings is still a challenge. Statistical Mechanics has long tackled computational problems, and this thesis applies physics-based insights to understand DNNs via three complementary approaches. First, by averaging over data, we derive an asymptotic bound on generalization that depends solely on the size of the last layer, rather than on the total number of parameters -- revealing how deep architectures process information differently across layers. Second, adopting a data-dependent viewpoint, we explore a finite-width thermodynamic limit beyond the infinite-width regime. This leads to: (i) a closed-form expression for the generalization error in a finite-width one-hidden-layer network (regression task); (ii) an approximate partition function for deeper architectures; and (iii) a link between deep networks in this thermodynamic limit and Student's t-processes. Finally, from a task-explicit perspective, we present a preliminary analysis of how DNNs interact with a controlled dataset, investigating whether they truly internalize its structure -- collapsing to the teacher -- or merely memorize it. By understanding when a network must learn data structure rather than just memorize, it sheds light on fostering meaningful internal representations. In essence, this thesis leverages the synergy between Statistical Physics and Machine Learning to illuminate the inner behavior of DNNs.


5 GANs Concepts You Should Know About in 2023 - MarkTechPost

#artificialintelligence

Generative modeling is an unsupervised learning task involving automatically discovering and learning the patterns in input data so that the model can generate new outputs that plausibly could have been drawn from the original dataset. GANs are generative models that can create new data points resembling the training data. For instance, GANs can produce pictures resembling photographs of human faces, even though the faces depicted do not correspond to any actual individual. GANs consist of two models – a generator and a discriminator. The discriminator is a Convolutional Neural Network (CNN) consisting of various hidden layers and one output layer. The generator is an Inverse Convolutional Neural Net doing exactly the opposite of what a CNN does because.


Pneumonia Detection in Chest X-Ray Images : Handling Class Imbalance

Ali, Wardah, Qureshi, Eesha, Farooqi, Omama Ahmed, Khan, Rizwan Ahmed

arXiv.org Artificial Intelligence

People all over the globe are affected by pneumonia but deaths due to it are highest in Sub-Saharan Asia and South Asia. In recent years, the overall incidence and mortality rate of pneumonia regardless of the utilization of effective vaccines and compelling antibiotics has escalated. Thus, pneumonia remains a disease that needs spry prevention and treatment. The widespread prevalence of pneumonia has caused the research community to come up with a framework that helps detect, diagnose and analyze diseases accurately and promptly. One of the major hurdles faced by the Artificial Intelligence (AI) research community is the lack of publicly available datasets for chest diseases, including pneumonia . Secondly, few of the available datasets are highly imbalanced (normal examples are over sampled, while samples with ailment are in severe minority) making the problem even more challenging. In this article we present a novel framework for the detection of pneumonia. The novelty of the proposed methodology lies in the tackling of class imbalance problem. The Generative Adversarial Network (GAN), specifically a combination of Deep Convolutional Generative Adversarial Network (DCGAN) and Wasserstein GAN gradient penalty (WGAN-GP) was applied on the minority class ``Pneumonia'' for augmentation, whereas Random Under-Sampling (RUS) was done on the majority class ``No Findings'' to deal with the imbalance problem. The ChestX-Ray8 dataset, one of the biggest datasets, is used to validate the performance of the proposed framework. The learning phase is completed using transfer learning on state-of-the-art deep learning models i.e. ResNet-50, Xception, and VGG-16. Results obtained exceed state-of-the-art.


#007 How to implement GAN Hacks to Train Stable Models?

#artificialintelligence

Highlights: In this post, we are going to learn several hacks that we can use to train stable GAN models. First, we are going to provide a quick recap of the GANs theory, and then, we are going to talk about challenges when training GANs. After that, we will provide solutions for these challenges in Python. So, let's begin with our post. Training Generative Adversarial Networks (GANs), can be quite a challenging task. This is mainly because two networks, discriminator and generator, have to be trained simultaneously.


Convolutional Neural Networks - AI Summary

#artificialintelligence

Research by Hubel and Wiesel [2,3] analyzed the striate cortex of cats and monkeys, revealing two key findings that would come to heavily influence Fukushima's work [1]. The next significant implementation of a convolution neural network was LeNet-5 proposed in 1999 by Le Cun et al. in their work "Object Recognition with Gradient Based Learning'' [4]. Their proposed network, LeNet-5 performed well on the MNIST data set and was shown to do better than state of the art (at the time) SVMs and K-nearest neighbor based approaches. Their final implementation outperformed other state of the art image classification algorithms with error rates which were 10% lower than its competitors on the ImageNet dataset. This application of a discrete convolution precisely represents local receptive fields observed by Hubel and Wiesel [2,3] and implemented in early CNNs by Fukushima and Le Cun [1,4]. Research by Hubel and Wiesel [2,3] analyzed the striate cortex of cats and monkeys, revealing two key findings that would come to heavily influence Fukushima's work [1]. The next significant implementation of a convolution neural network was LeNet-5 proposed in 1999 by Le Cun et al. in their work "Object Recognition with Gradient Based Learning'' [4].


Using GANs to Create Anime Faces via Pytorch

#artificialintelligence

Most of us in data science have seen a lot of AI-generated people in recent times, whether it be in papers, blogs, or videos. We've reached a stage where it's becoming increasingly difficult to distinguish between actual human faces and faces generated by artificial intelligence. However, with the current available machine learning toolkits, creating these images yourself is not as difficult as you might think. In my view, GANs will change the way we generate video games and special effects. Using this approach, we could create realistic textures or characters on demand.


Evaluation of Deep Convolutional Generative Adversarial Networks for data augmentation of chest X-ray images

Venu, Sagar Kora

arXiv.org Artificial Intelligence

Medical image datasets are usually imbalanced, due to the high costs of obtaining the data and time-consuming annotations. Training deep neural network models on such datasets to accurately classify the medical condition does not yield desired results and often over-fits the data on majority class samples. In order to address this issue, data augmentation is often performed on training data by position augmentation techniques such as scaling, cropping, flipping, padding, rotation, translation, affine transformation, and color augmentation techniques such as brightness, contrast, saturation, and hue to increase the dataset sizes. These augmentation techniques are not guaranteed to be advantageous in domains with limited data, especially medical image data, and could lead to further overfitting. In this work, we performed data augmentation on the Chest X-rays dataset through generative modeling (deep convolutional generative adversarial network) which creates artificial instances retaining similar characteristics to the original data and evaluation of the model resulted in Fr\'echet Distance of Inception (FID) score of 1.289.


Deep convolutional generative adversarial networks for traffic data imputation encoding time series as images

Huang, Tongge, Chakraborty, Pranamesh, Sharma, Anuj

arXiv.org Artificial Intelligence

Sufficient high-quality traffic data are a crucial component of various Intelligent Transportation System (ITS) applications and research related to congestion prediction, speed prediction, incident detection, and other traffic operation tasks. Nonetheless, missing traffic data are a common issue in sensor data which is inevitable due to several reasons, such as malfunctioning, poor maintenance or calibration, and intermittent communications. Such missing data issues often make data analysis and decision-making complicated and challenging. In this study, we have developed a generative adversarial network (GAN) based traffic sensor data imputation framework (TSDIGAN) to efficiently reconstruct the missing data by generating realistic synthetic data. In recent years, GANs have shown impressive success in image data generation. However, generating traffic data by taking advantage of GAN based modeling is a challenging task, since traffic data have strong time dependency. To address this problem, we propose a novel time-dependent encoding method called the Gramian Angular Summation Field (GASF) that converts the problem of traffic time-series data generation into that of image generation. We have evaluated and tested our proposed model using the benchmark dataset provided by Caltrans Performance Management Systems (PeMS). This study shows that the proposed model can significantly improve the traffic data imputation accuracy in terms of Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE) compared to state-of-the-art models on the benchmark dataset. Further, the model achieves reasonably high accuracy in imputation tasks even under a very high missing data rate ($>$ 50\%), which shows the robustness and efficiency of the proposed model.


Deep Learning - Deep Convolutional Generative Adversarial Networks Basics Vinod Sharma's Blog

#artificialintelligence

I introduced the basic analogy, concept, and ideas behind "How GANs work". This post will do a little bit of a deep dive. Generative Adversarial Networks are a class of algorithms used in the unsupervised learning environment. As the name suggests they are called Adversarial Networks because they are is made up of two competing neural networks. Both networks compete with each other to achieve a zero-sum game. Both neural networks are assigned different job role i.e. contesting with each other. The process in GANs involves automatically learning to discover the regularities or patterns in input data.


The Machine Learning (ML) Bootcamp

#artificialintelligence

Get your team access to 3,500 top Udemy courses anytime, anywhere. Get your team access to 3,500 top Udemy courses anytime, anywhere. Maths: Calculus, Linear Algebra, Statistics, Naive Bayes Methods: Neural Networks, Deep Learning, PCA, Scikit-learn, Tensorflow, Keras Machine: Python, Cloud Computing, Colab Insights into real life projects and how to apply the concepts Do you want to master Machine Learning (ML) - the key field of the future? ML is the core of artificial intelligence and will transform all industries and all areas of life. This comprehensive course covers the three M's Maths, Methods and Machine, and is easy to understand.